Goto

Collaborating Authors

 complementary label


Learning from Complementary Labels

Neural Information Processing Systems

Collecting labeled data is costly and thus a critical bottleneck in real-world classification tasks. To mitigate this problem, we propose a novel setting, namely learning from complementary labels for multi-class classification. A complementary label specifies a class that a pattern does not belong to. Collecting complementary labels would be less laborious than collecting ordinary labels, since users do not have to carefully choose the correct class from a long list of candidate classes. However, complementary labels are less informative than ordinary labels and thus a suitable approach is needed to better learn from them. In this paper, we show that an unbiased estimator to the classification risk can be obtained only from complementarily labeled data, if a loss function satisfies a particular symmetric condition. We derive estimation error bounds for the proposed method and prove that the optimal parametric convergence rate is achieved. We further show that learning from complementary labels can be easily combined with learning from ordinary labels (i.e., ordinary supervised learning), providing a highly practical implementation of the proposed method. Finally, we experimentally demonstrate the usefulness of the proposed methods.



Provably Consistent Partial-Label Learning: Supplementary Material A Proofs of Data Generation Process A.1 Proof of Theorem 1

Neural Information Processing Systems

A.3 Proof of Lemma 1 Let us first consider the case where the correct label y is a specific label i ( i [ k ]), then we have p(y Y,y = i | x) = p( y Y | y = i, x)p (y = i | x) = null Our proof of the estimation error bound is based on Rademacher complexity [1]. Before proving Theorem 4, we introduce the following lemmas. The same proof has been provided in [20]. Then we have the following lemma. Since this proof is somewhat similar to the proof of Theorem 4, we briefly sketch the key points.


Appendix A Proofs A.1 Proof of Proposition

Neural Information Processing Systems

For the general backward correction, based on Eq. 11, conducting adversarial training (A T) on the Based on Eqs. 13, 14 and 15, the inequality holds between their empirical formulations: We adversarially train a model with several complementary losses separately on Kuzushiji. The results show the same observation (as in Section 4.2) Note that we only optimize the model using the ones generated by the oracle. Figure 6: The results on four randomly sampled instances from Kuzushiji. A T with CLs, the two-stage method consists of a complementary learning phase and an A T phase, following the setups of complementary learning setups and A T setups (in Section 5), respectively. For CIFAR10 and SVHN, their learning rates are set to 0.01.


Adversarial Training with Complementary Labels: On the Benefit of Gradually Informative Attacks

Neural Information Processing Systems

To push A T towards more practical scenarios, we explore a brand new yet challenging setting, i.e., A T with complementary labels (CLs), which specify a class that a data sample does not belong to.



Scalable Oversight via Partitioned Human Supervision

Yin, Ren, Ishida, Takashi, Sugiyama, Masashi

arXiv.org Artificial Intelligence

As artificial intelligence (AI) systems approach and surpass expert human performance across a broad range of tasks, obtaining high-quality human supervision for evaluation and training becomes increasingly challenging. Our focus is on tasks that require deep knowledge and skills of multiple domains. Unfortunately, even the best human experts are knowledgeable only in a single narrow area, and will not be able to evaluate the correctness of advanced AI systems on such superhuman tasks. However, based on their narrow expertise, humans may provide a weak signal, i.e., a complementary label indicating an option that is incorrect. For example, a cardiologist could state that "this is not related to cardiology,'' even if they cannot identify the true disease. Based on this weak signal, we propose a scalable oversight framework that enables us to evaluate frontier AI systems without the need to prepare the ground truth. We derive an unbiased estimator of top-1 accuracy from complementary labels and quantify how many complementary labels are needed to match the variance of ordinary labels. We further introduce two estimators to combine scarce ordinary labels with abundant complementary labels. We provide finite-sample deviation guarantees for both complementary-only and the mixed estimators. Empirically, we show that we can evaluate the output of large language models without the ground truth, if we have complementary labels. We further show that we can train an AI system with such weak signals: we show how we can design an agentic AI system automatically that can perform better with this partitioned human supervision. Our code is available at https://github.com/R-Yin-217/Scalable-Oversight-via-Human-Partitioned-Supervision.


Rethinking Consistent Multi-Label Classification under Inexact Supervision

Wang, Wei, Ma, Tianhao, Xie, Ming-Kun, Niu, Gang, Sugiyama, Masashi

arXiv.org Artificial Intelligence

Partial multi-label learning and complementary multi-label learning are two popular weakly supervised multi-label classification paradigms that aim to alleviate the high annotation costs of collecting precisely annotated multi-label data. In partial multi-label learning, each instance is annotated with a candidate label set, among which only some labels are relevant; in complementary multi-label learning, each instance is annotated with complementary labels indicating the classes to which the instance does not belong. Existing consistent approaches for the two paradigms either require accurate estimation of the generation process of candidate or complementary labels or assume a uniform distribution to eliminate the estimation problem. However, both conditions are usually difficult to satisfy in real-world scenarios. In this paper, we propose consistent approaches that do not rely on the aforementioned conditions to handle both problems in a unified way. Specifically, we propose two unbiased risk estimators based on first- and second-order strategies. Theoretically, we prove consistency w.r.t. two widely used multi-label classification evaluation metrics and derive convergence rates for the estimation errors of the proposed risk estimators. Empirically, extensive experimental results validate the effectiveness of our proposed approaches against state-of-the-art methods.